Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 8 de 8
Filtrar
Mais filtros










Base de dados
Intervalo de ano de publicação
1.
Clin Linguist Phon ; 35(3): 253-276, 2021 03 04.
Artigo em Inglês | MEDLINE | ID: mdl-32567986

RESUMO

Recent studies on the remediation of speech disorders suggest that providing visual information of speech articulators may contribute to improve speech production. In this study, we evaluate the effectiveness of an illustration-based rehabilitation method on speech recovery of a patient with non-fluent chronic aphasia. The Ultraspeech-player software allowed visualization by the patient of reference tongue and lip movements recorded using ultrasound and video imaging. This method can improve the patient's awareness of their own lingual and labial movements, which can increase the ability to coordinate and combine articulatory gestures. The effects of this method were assessed by analyzing performance during speech tasks, the phonological processes identified in the errors made during the phoneme repetition task and the acoustic parameters derived from the speech signal. We also evaluated cognitive performance before and after rehabilitation. The integrity of visuospatial ability, short-term and working memory and some executive functions supports the effectiveness of the rehabilitation method. Our results showed that illustration-based rehabilitation technique had a beneficial effect on the patient's speech production, especially for stop and fricative consonants which are targeted (high visibility of speech articulator configurations) by the software, but also on reading abilities. Acoustic parameters indicated an improvement in the distinction between consonant categories: voiced and voiceless stops or alveolar, post-alveolar and labiodental fricatives. However, the patient showed little improvement for vowels. These results confirmed the advantage of using illustration-based rehabilitation technique and the necessity of detailed subjective and objective intra-speaker evaluation in speech production to fully evaluate speech abilities.


Assuntos
Afasia , Articuladores Dentários , Humanos , Fonética , Fala , Medida da Produção da Fala , Fonoterapia
2.
Clin Linguist Phon ; 34(9): 826-843, 2020 09 01.
Artigo em Inglês | MEDLINE | ID: mdl-31992079

RESUMO

Intraoral surgery for tongue cancer usually induces speech disorders that have a negative impact on communication and quality of life. Studies have documented the benefit of tongue ultrasound imaging as a visual articulatory feedback for speech rehabilitation. This study aims to assess specifically the complementary contribution of visual feedback to visual illustration (i.e. the display of ultrasound video of target language movements) for the speech rehabilitation of glossectomised patients. Two therapy conditions were used alternately for ten glossectomised French patients randomly divided into two cohorts. The IF cohort benefitted from 10 sessions using illustration alone (IL condition) followed by 10 sessions using illustration supplemented by visual feedback (IL+F condition). The FI cohort followed the opposite protocol, i.e. the first 10 sessions with the IL+F condition, followed by 10 sessions with the IL condition. Phonetic accuracy (Percent Consonants Correct) was monitored at baseline (T0, before the first series) and after each series (T1 and T2) using clinical speech-language assessments. None of the contrasts computed between the two conditions, using logistic regression with random effects models, were found to be statistically significant for the group analysis of assessment scores. Results were significant for a few individuals, with balanced advantages in both conditions. In conclusion, the use of articulatory visual feedback does not seem to bring a decisive advantage over the use of visual illustration, though speech therapists and patients reported that ultrasound feedback was useful at the beginning. This result should be confirmed by similar studies involving other types of speech disorders.


Assuntos
Retroalimentação Sensorial , Retroalimentação , Glossectomia , Distúrbios da Fala/reabilitação , Ultrassonografia , Feminino , França , Humanos , Idioma , Masculino , Pessoa de Meia-Idade , Fonética , Língua/diagnóstico por imagem , Neoplasias da Língua/cirurgia
3.
Neural Comput ; 32(3): 596-625, 2020 03.
Artigo em Inglês | MEDLINE | ID: mdl-31951798

RESUMO

Sensory processing is increasingly conceived in a predictive framework in which neurons would constantly process the error signal resulting from the comparison of expected and observed stimuli. Surprisingly, few data exist on the accuracy of predictions that can be computed in real sensory scenes. Here, we focus on the sensory processing of auditory and audiovisual speech. We propose a set of computational models based on artificial neural networks (mixing deep feedforward and convolutional networks), which are trained to predict future audio observations from present and past audio or audiovisual observations (i.e., including lip movements). Those predictions exploit purely local phonetic regularities with no explicit call to higher linguistic levels. Experiments are conducted on the multispeaker LibriSpeech audio speech database (around 100 hours) and on the NTCD-TIMIT audiovisual speech database (around 7 hours). They appear to be efficient in a short temporal range (25-50 ms), predicting 50% to 75% of the variance of the incoming stimulus, which could result in potentially saving up to three-quarters of the processing power. Then they quickly decrease and almost vanish after 250 ms. Adding information on the lips slightly improves predictions, with a 5% to 10% increase in explained variance. Interestingly the visual gain vanishes more slowly, and the gain is maximum for a delay of 75 ms between image and predicted sound.


Assuntos
Percepção Auditiva/fisiologia , Aprendizado Profundo , Fala/fisiologia , Estimulação Acústica , Humanos , Redes Neurais de Computação , Estimulação Luminosa , Percepção Visual/fisiologia
4.
Clin Linguist Phon ; 32(7): 595-621, 2018.
Artigo em Inglês | MEDLINE | ID: mdl-29148845

RESUMO

The rehabilitation of speech disorders benefits from providing visual information which may improve speech motor plans in patients. We tested the proof of concept of a rehabilitation method (Sensori-Motor Fusion, SMF; Ultraspeech player) in one post-stroke patient presenting chronic non-fluent aphasia. SMF allows visualisation by the patient of target tongue and lips movements using high-speed ultrasound and video imaging. This can improve the patient's awareness of his/her own lingual and labial movements, which can, in turn, improve the representation of articulatory movements and increase the ability to coordinate and combine articulatory gestures. The auditory and oro-sensory feedback received by the patient as a result of his/her own pronunciation can be integrated with the target articulatory movements they watch. Thus, this method is founded on sensorimotor integration during speech. The SMF effect on this patient was assessed through qualitative comparison of language scores and quantitative analysis of acoustic parameters measured in a speech production task, before and after rehabilitation. We also investigated cerebral patterns of language reorganisation for rhyme detection and syllable repetition, to evaluate the influence of SMF on phonological-phonetic processes. Our results showed that SMF had a beneficial effect on this patient who qualitatively improved in naming, reading, word repetition and rhyme judgment tasks. Quantitative measurements of acoustic parameters indicate that the patient's production of vowels and syllables also improved. Compared with pre-SMF, the fMRI data in the post-SMF session revealed the activation of cerebral regions related to articulatory, auditory and somatosensory processes, which were expected to be recruited by SMF. We discuss neurocognitive and linguistic mechanisms which may explain speech improvement after SMF, as well as the advantages of using this speech rehabilitation method.


Assuntos
Afasia de Broca/terapia , Idioma , Plasticidade Neuronal , Fonoterapia/métodos , Fala/fisiologia , Retroalimentação Sensorial/fisiologia , Feminino , Humanos , Lábio , Imageamento por Ressonância Magnética , Língua
5.
Neuropsychologia ; 109: 126-133, 2018 01 31.
Artigo em Inglês | MEDLINE | ID: mdl-29248497

RESUMO

Recent neurophysiological studies demonstrate that audio-visual speech integration partly operates through temporal expectations and speech-specific predictions. From these results, one common view is that the binding of auditory and visual, lipread, speech cues relies on their joint probability and prior associative audio-visual experience. The present EEG study examined whether visual tongue movements integrate with relevant speech sounds, despite little associative audio-visual experience between the two modalities. A second objective was to determine possible similarities and differences of audio-visual speech integration between unusual audio-visuo-lingual and classical audio-visuo-labial modalities. To this aim, participants were presented with auditory, visual, and audio-visual isolated syllables, with the visual presentation related to either a sagittal view of the tongue movements or a facial view of the lip movements of a speaker, with lingual and facial movements previously recorded by an ultrasound imaging system and a video camera. In line with previous EEG studies, our results revealed an amplitude decrease and a latency facilitation of P2 auditory evoked potentials in both audio-visual-lingual and audio-visuo-labial conditions compared to the sum of unimodal conditions. These results argue against the view that auditory and visual speech cues solely integrate based on prior associative audio-visual perceptual experience. Rather, they suggest that dynamic and phonetic informational cues are sharable across sensory modalities, possibly through a cross-modal transfer of implicit articulatory motor knowledge.


Assuntos
Encéfalo/fisiologia , Percepção de Movimento/fisiologia , Percepção da Fala/fisiologia , Língua , Adulto , Eletroencefalografia , Potenciais Evocados , Retroalimentação , Feminino , Humanos , Lábio , Leitura Labial , Masculino , Pessoa de Meia-Idade , Reconhecimento Visual de Modelos/fisiologia , Percepção Social , Adulto Jovem
6.
J Cogn Neurosci ; 29(3): 448-466, 2017 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-28139959

RESUMO

Action recognition has been found to rely not only on sensory brain areas but also partly on the observer's motor system. However, whether distinct auditory and visual experiences of an action modulate sensorimotor activity remains largely unknown. In the present sparse sampling fMRI study, we determined to which extent sensory and motor representations interact during the perception of tongue and lip speech actions. Tongue and lip speech actions were selected because tongue movements of our interlocutor are accessible via their impact on speech acoustics but not visible because of its position inside the vocal tract, whereas lip movements are both "audible" and visible. Participants were presented with auditory, visual, and audiovisual speech actions, with the visual inputs related to either a sagittal view of the tongue movements or a facial view of the lip movements of a speaker, previously recorded by an ultrasound imaging system and a video camera. Although the neural networks involved in visual visuolingual and visuofacial perception largely overlapped, stronger motor and somatosensory activations were observed during visuolingual perception. In contrast, stronger activity was found in auditory and visual cortices during visuofacial perception. Complementing these findings, activity in the left premotor cortex and in visual brain areas was found to correlate with visual recognition scores observed for visuolingual and visuofacial speech stimuli, respectively, whereas visual activity correlated with RTs for both stimuli. These results suggest that unimodal and multimodal processing of lip and tongue speech actions rely on common sensorimotor brain areas. They also suggest that visual processing of audible but not visible movements induces motor and visual mental simulation of the perceived actions to facilitate recognition and/or to learn the association between auditory and visual signals.


Assuntos
Encéfalo/fisiologia , Reconhecimento Facial/fisiologia , Percepção de Movimento/fisiologia , Percepção da Fala/fisiologia , Estimulação Acústica/métodos , Adolescente , Adulto , Encéfalo/diagnóstico por imagem , Mapeamento Encefálico , Feminino , Humanos , Imageamento por Ressonância Magnética , Masculino , Testes Neuropsicológicos , Estimulação Luminosa/métodos , Tempo de Reação , Percepção Social , Adulto Jovem
7.
PLoS Comput Biol ; 12(11): e1005119, 2016 Nov.
Artigo em Inglês | MEDLINE | ID: mdl-27880768

RESUMO

Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips) into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN) trained on electromagnetic articulography (EMA) data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer.


Assuntos
Biorretroalimentação Psicológica/métodos , Interfaces Cérebro-Computador , Auxiliares de Comunicação para Pessoas com Deficiência , Redes Neurais de Computação , Espectrografia do Som/métodos , Medida da Produção da Fala/métodos , Biorretroalimentação Psicológica/instrumentação , Sistemas Computacionais , Humanos , Fonética , Espectrografia do Som/instrumentação , Acústica da Fala , Inteligibilidade da Fala , Medida da Produção da Fala/instrumentação
8.
J Physiol Paris ; 110(4 Pt A): 392-401, 2016 11.
Artigo em Inglês | MEDLINE | ID: mdl-28756027

RESUMO

Restoring communication in case of aphasia is a key challenge for neurotechnologies. To this end, brain-computer strategies can be envisioned to allow artificial speech synthesis from the continuous decoding of neural signals underlying speech imagination. Such speech brain-computer interfaces do not exist yet and their design should consider three key choices that need to be made: the choice of appropriate brain regions to record neural activity from, the choice of an appropriate recording technique, and the choice of a neural decoding scheme in association with an appropriate speech synthesis method. These key considerations are discussed here in light of (1) the current understanding of the functional neuroanatomy of cortical areas underlying overt and covert speech production, (2) the available literature making use of a variety of brain recording techniques to better characterize and address the challenge of decoding cortical speech signals, and (3) the different speech synthesis approaches that can be considered depending on the level of speech representation (phonetic, acoustic or articulatory) envisioned to be decoded at the core of a speech BCI paradigm.


Assuntos
Interfaces Cérebro-Computador , Fala/fisiologia , Humanos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...